深度神经网络需要特定的层来处理点云,因为点的分散和不规则位置使我们无法使用卷积过滤器。在这里,我们介绍了复合层,该复合层是点云的新卷积操作员。我们的复合层的特殊性是,它在将点与其特征向量结合之前从点位置提取和压缩空间信息。与众所周知的点横向跨层相比,我们的复合层提供了额外的正则化,并确保了参数和参数数量方面的灵活性更大。为了展示设计灵活性,我们还定义了一个集合复合层,该复合层以非线性方式组合空间信息和特征,并且我们使用这些层来实现卷积和聚集的综合材料。我们训练我们的复合烯类进行分类,最引人注目的是无监督的异常检测。我们对合成和现实世界数据集的实验表明,在这两个任务中,我们的CompositeNets都优于表现要点,尽管具有更简单的体系结构,但取得了与KPCONV相似的结果。此外,我们的复合烯类基本上优于现有的解决方案,用于点云上的异常检测。
translated by 谷歌翻译
任何电子设备中包含的芯片都是通过圆形硅晶片制造的,这些芯片是通过不同生产阶段的检查机对其进行监控的。检查机检测并找到晶圆中的任何缺陷,并返回晶圆缺陷图(WDM),即,缺陷为lie的坐标列表,可以将其视为巨大,稀疏和二进制图像。在正常情况下,晶片表现出少量随机分布的缺陷,而以特定模式分组的缺陷可能表明生产线中的已知或新颖类别。不用说,半导体行业的主要关注点是确定这些模式并尽快进行干预以恢复正常的生产条件。在这里,我们将WDM监视作为开放式识别问题,以准确地将WDM分类为已知类别并迅速检测到新颖的模式。特别是,我们提出了一条基于Submanifold稀疏卷积网络的晶圆监测的综合管道,这是一种深层体系结构,旨在以任意分辨率处理稀疏数据,并在已知类别上进行了培训。为了检测新颖性,我们根据拟合在分类器潜在表示上的高斯混合模型定义了一个离群检测器。我们在WDM的真实数据集上进行的实验表明,Submanifold稀疏卷积直接处​​理全分辨率WDMS在已知类别上比传统的卷积神经网络产生了卓越的分类性能,这需要初步的封装以减少代表WDM的二元图像的大小。此外,我们的解决方案优于最先进的开放式识别解决方案,以检测新颖性。
translated by 谷歌翻译
我们解决了多变量数据流中的在线变更检测问题,并介绍了Quanttree指数加权移动平均值(QT-EWMA),这是一种非参数变更检测算法,可以在误报之前控制预期的时间,从Arl $ _0 $)。在许多应用程序中,控制虚假警报至关重要,很少能通过在线变更检测算法来保证,这些算法可以监视多元数据串联而不知道数据分布。像许多变更检测算法一样,QT-EWMA从固定训练集中构建了数据分布的模型,在我们的情况下,量化量子三直方图。为了监视数据流,即使训练集非常小,我们提出了QT-Ewma-update,该QT-ewma-update在监视过程中会逐步更新Quanttree直方图,请始终保持ARL $ _0 $的控制。我们的实验在合成和真实的数据源上执行,证明了QT-Ewma和Qt-Ewma-update控制ARL $ _0 $和错误警报率比在类似条件下运行的最先进方法更好,从而实现了错误的警报率。较低或可比的检测延迟。
translated by 谷歌翻译
Numerous works use word embedding-based metrics to quantify societal biases and stereotypes in texts. Recent studies have found that word embeddings can capture semantic similarity but may be affected by word frequency. In this work we study the effect of frequency when measuring female vs. male gender bias with word embedding-based bias quantification methods. We find that Skip-gram with negative sampling and GloVe tend to detect male bias in high frequency words, while GloVe tends to return female bias in low frequency words. We show these behaviors still exist when words are randomly shuffled. This proves that the frequency-based effect observed in unshuffled corpora stems from properties of the metric rather than from word associations. The effect is spurious and problematic since bias metrics should depend exclusively on word co-occurrences and not individual word frequencies. Finally, we compare these results with the ones obtained with an alternative metric based on Pointwise Mutual Information. We find that this metric does not show a clear dependence on frequency, even though it is slightly skewed towards male bias across all frequencies.
translated by 谷歌翻译
This report summarizes the work carried out by the authors during the Twelfth Montreal Industrial Problem Solving Workshop, held at Universit\'e de Montr\'eal in August 2022. The team tackled a problem submitted by CBC/Radio-Canada on the theme of Automatic Text Simplification (ATS).
translated by 谷歌翻译
Feature acquisition algorithms address the problem of acquiring informative features while balancing the costs of acquisition to improve the learning performances of ML models. Previous approaches have focused on calculating the expected utility values of features to determine the acquisition sequences. Other approaches formulated the problem as a Markov Decision Process (MDP) and applied reinforcement learning based algorithms. In comparison to previous approaches, we focus on 1) formulating the feature acquisition problem as a MDP and applying Monte Carlo Tree Search, 2) calculating the intermediary rewards for each acquisition step based on model improvements and acquisition costs and 3) simultaneously optimizing model improvement and acquisition costs with multi-objective Monte Carlo Tree Search. With Proximal Policy Optimization and Deep Q-Network algorithms as benchmark, we show the effectiveness of our proposed approach with experimental study.
translated by 谷歌翻译
360-degree panoramic videos have gained considerable attention in recent years due to the rapid development of head-mounted displays (HMDs) and panoramic cameras. One major problem in streaming panoramic videos is that panoramic videos are much larger in size compared to traditional ones. Moreover, the user devices are often in a wireless environment, with limited battery, computation power, and bandwidth. To reduce resource consumption, researchers have proposed ways to predict the users' viewports so that only part of the entire video needs to be transmitted from the server. However, the robustness of such prediction approaches has been overlooked in the literature: it is usually assumed that only a few models, pre-trained on past users' experiences, are applied for prediction to all users. We observe that those pre-trained models can perform poorly for some users because they might have drastically different behaviors from the majority, and the pre-trained models cannot capture the features in unseen videos. In this work, we propose a novel meta learning based viewport prediction paradigm to alleviate the worst prediction performance and ensure the robustness of viewport prediction. This paradigm uses two machine learning models, where the first model predicts the viewing direction, and the second model predicts the minimum video prefetch size that can include the actual viewport. We first train two meta models so that they are sensitive to new training data, and then quickly adapt them to users while they are watching the videos. Evaluation results reveal that the meta models can adapt quickly to each user, and can significantly increase the prediction accuracy, especially for the worst-performing predictions.
translated by 谷歌翻译
This paper presents a corpus annotated for the task of direct-speech extraction in Croatian. The paper focuses on the annotation of the quotation, co-reference resolution, and sentiment annotation in SETimes news corpus in Croatian and on the analysis of its language-specific differences compared to English. From this, a list of the phenomena that require special attention when performing these annotations is derived. The generated corpus with quotation features annotations can be used for multiple tasks in the field of Natural Language Processing.
translated by 谷歌翻译
With the ever-growing popularity of the field of NLP, the demand for datasets in low resourced-languages follows suit. Following a previously established framework, in this paper, we present the UNER dataset, a multilingual and hierarchical parallel corpus annotated for named-entities. We describe in detail the developed procedure necessary to create this type of dataset in any language available on Wikipedia with DBpedia information. The three-step procedure extracts entities from Wikipedia articles, links them to DBpedia, and maps the DBpedia sets of classes to the UNER labels. This is followed by a post-processing procedure that significantly increases the number of identified entities in the final results. The paper concludes with a statistical and qualitative analysis of the resulting dataset.
translated by 谷歌翻译
This article presents the application of the Universal Named Entity framework to generate automatically annotated corpora. By using a workflow that extracts Wikipedia data and meta-data and DBpedia information, we generated an English dataset which is described and evaluated. Furthermore, we conducted a set of experiments to improve the annotations in terms of precision, recall, and F1-measure. The final dataset is available and the established workflow can be applied to any language with existing Wikipedia and DBpedia. As part of future research, we intend to continue improving the annotation process and extend it to other languages.
translated by 谷歌翻译